Proximal Backpropagation

نویسندگان

  • Thomas Frerix
  • Thomas Möllenhoff
  • Michael Möller
  • Daniel Cremers
چکیده

We propose proximal backpropagation (ProxProp) as a novel algorithm that takes implicit instead of explicit gradient steps to update the network parameters during neural network training. Our algorithm is motivated by the step size limitation of explicit gradient descent, which poses an impediment for optimization. ProxProp is developed from a general point of view on the backpropagation algorithm, currently the most common technique to train neural networks via stochastic gradient descent and variants thereof. Specifically, we show that backpropagation of a prediction error is equivalent to sequential gradient descent steps on a quadratic penalty energy, which comprises the network activations as variables of the optimization. We further analyze theoretical properties of ProxProp and in particular prove that the algorithm yields a descent direction in parameter space and can therefore be combined with a wide variety of convergent algorithms. Finally, we devise an efficient numerical implementation that integrates well with popular deep learning frameworks. We conclude by demonstrating promising numerical results and show that ProxProp can be effectively combined with common first order optimizers such as Adam.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Ca2+ imaging of mouse neocortical interneurone dendrites: Ia-type K+ channels control action potential backpropagation.

GABAergic interneurones are essential in cortical processing, yet the functional properties of their dendrites are still poorly understood. In this first study, we combined two-photon calcium imaging with whole-cell recording and anatomical reconstructions to examine the calcium dynamics during action potential (AP) backpropagation in three types of V1 supragranular interneurones: parvalbumin-p...

متن کامل

Dichotomy of action-potential backpropagation in CA1 pyramidal neuron dendrites.

In hippocampal CA1 pyramidal neurons, action potentials are typically initiated in the axon and backpropagate into the dendrites, shaping the integration of synaptic activity and influencing the induction of synaptic plasticity. Despite previous reports describing action-potential propagation in the proximal apical dendrites, the extent to which action potentials invade the distal dendrites of ...

متن کامل

Investigation of Mechanical Properties of Self Compacting Polymeric Concrete with Backpropagation Network

Acrylic polymer that is highly stable against chemicals and is a good choice when concrete is subject to chemical attack. In this study, self-compacting concrete (SCC) made using acrylic polymer, nanosilica and microsilica has been investigated. The results of experimental testing showed that the addition of microsilica and acrylic polymer decreased the tensile, compressive and bending strength...

متن کامل

Npgrj_nn_2040 178..186

The axon initial segment (AIS) is a specialized region in neurons where action potentials are initiated. It is commonly assumed that this process requires a high density of voltage-gated sodium (Na1) channels. Paradoxically, the results of patch-clamp studies suggest that the Na1 channel density at the AIS is similar to that at the soma and proximal dendrites. Here we provide data obtained by a...

متن کامل

Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-\L{}ojasiewicz Condition

In 1963, Polyak proposed a simple condition that is sufficient to show a global linear convergence rate for gradient descent. This condition is a special case of the Lojasiewicz inequality proposed in the same year, and it does not require strong convexity (or even convexity). In this work, we show that this much-older PolyakLojasiewicz (PL) inequality is actually weaker than the main condition...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1706.04638  شماره 

صفحات  -

تاریخ انتشار 2017